Web Survey Bibliography
In survey research, computer assisted forms of data collection are rapidly replacing paper and pencil methods. At the same time children are becoming important respondents in many large-scale surveys. However, children possess distinctive cognitive and social developmental characteristics, what must be taken into consideration in the research design. There are many problems to be solved when the respondents are children, including problems of language use, literacy and different stages of cognitive development (Scott 2001).
In order to examine data quality collected with computer assisted self administered questionnaires (CASQ) a study among 9 and 10 years old children was conducted. Within the study two hypotheses were tested:
1. “Cognitive skills do have an impact on reliability of data collected by CASQ, but reliability is not lower than in paper and pencil collection mode.”
2. “Computer skills do not have a significant influence on quality of data in CSAQ.”
In order to answer research questions three different databases were used – PIRLS 2001 (Progress in International Reading Literacy Study) (n=3118); TIMSS 2003 (Trends in International Mathematics and Science Study) (n=3126) – both paper and pencil self administered studies, and computer assisted self administered study where several questions from paper and pencil studies were repeated with the intention to compare the results (n=150). All three studies were conducted in a school setting.
Four constructs were analyzed in relation to the following variables: mode of data collection, index of cognitive level, gender and index of computer skills.
The results show that cognitively more sophisticated children give somehow more reliable results, although further analysis showed the differences between groups are not statistically significant.
Reliability analyses were performed to compare data quality of groups of respondents with high and low index of computer use. Although Cronbach’s α is higher for group of respondents who use computers rarely, the difference between two groups is not statistically significant. Missing value analyses also shows, that respondents who use computers more rarely produced less item nonresponse compared to respondents who use computers more often, meaning that children with lower index of computer skills do not produce lower quality data.
General online research (GOR) 2008 (abstract)
Web survey bibliography - Conference proceedings (83)
- Estimation and Adjustment of Self-Selection Bias in Volunteer Panel Web Surveys ; 2016; Niu, Ch.
- Shorter Interviews, Longer Surveys: Optimising the survey participant experience whilst accommodating...; 2016; Halder, A.; Bansal, H. S.; Knowles, R.; Eldridge, J.; Murray, Mi.
- Gamifying. Not all fun and games; 2016; Stubington, P.; Crichton, C.
- Are interviews costing £0.08 a waste of money? Reviewing Google Surveys for Wisdom of the Crowd...; 2016; Roughton, G.; MacKay, I.
- Observations from Twelve Years of an Annual Market Research Technology Survey; 2016; Macer, T.; Wilson, S.
- A Comparison of the Effects of Face-to-Face and Online Deliberation on Young Students’ Attitudes...; 2015; Triantafillidou, A.; Yannas, P.; Lappas, G.; Kleftodimos, A.
- A Privacy-Friendly Method to Reward Participants of Online-Surveys; 2015; Herfert, M.; Lange, B.; Selzer, A.; Waldmann, U.
- Designing Bonsai Surveys: The small but perfectly formed survey experience to meet the needs of the...; 2015; Puleston, J.
- Is accuracy only for probability samples? Comparing probability and non-probability samples in a country...; 2013; Martinsson, J., Dahlberg, S., Lundmark, S.
- The effect of language in answering qualitative questions in user experience evaluation web-surveys; 2013; Walsh, T., Nurkka, P., Petrie, H., Olson, J.
- Beyond Satisfaction Questionnaires: “Hacking” the Online Survey; 2013; Evans, A. L.
- Advancing the field of questionnaire translation - identifying problems, discussing methods, pushing...; 2013; Behr, D., Dorer, B., Van Houten, G
- European Values Study - methodological and substantive applications; 2013; Luijkx, R., Jagodzinski, W.
- The Impact of Culture and Economy on Values and Attitudes; 2013; Duelmer, H., Voicu, M.
- Educational attainment in cross-national surveys: instrument design, data collection, harmonisation...; 2013; Schneider, S.
- Mode Effects in Mixed-Mode Surveys: Prevention, Diagnostics, and Adjustment 1; 2013; de Leeuw, E. D., Dillman, D. A., Schouten, B.
- The smart(phone) way to collect survey data; 2013; Stapleton, C.
- Unintentional mobile respondents; 2012; Peterson, G.
- Metering mobile usage. Insights from global Arbitron mobile trends panel; 2012; Verkasalo, H.
- Is „chapterisation“ a viable alternative to traditional progress indicators ?; 2012; Spicer, R., Dowling, Z.
- Self-administered mobile surveys; 2011; Bosnjak, M.
- Online survey research: Findings, Best practices, and future research; 2011
- Blend, balance, and stabilize respondent sources; 2011; Eggers, M., Drake, E.
- Mode Effect or Question Wording? Measurement Error in Mixed Mode Surveys; 2011; de Leeuw, E. D., Hox, J., Scherpenzeel, A.
- There is an app for that! A review of smartphone apps for marketing research; 2010; Michelson, M.
- The state of online research in the U.S.; 2010; Miller, J.
- A framework for understanding and applying ethical principles in network and security research; 2010; Kenneally, E., Bailey, M., Maughan, D.
- Restructuring and innovations on the survey “capacity of collective tourist accommodation”...; 2010; Santoro, M. T., Staffieri, S.
- An Analyze of the Zero Price Effect on Online Business Performance - An Research Based on the Mobile...; 2010; Liu, Y., Yuan, P.
- Dealing with Nonresponse in Survey Sampling: an Item Response Modeling Approach; 2010; Matei, A.
- Response format effects on measurement of employment; 2009; Thomas, R. K., Dillman, D. A., Smyth, J. D.
- Response Mode and Bias Analysis in the IRS’ Individual Taxpayer Burden Survey; 2009; Brick, J. M., Contos, G., Masken, K., Nord, R.
- Survey Mode Effects in Two Military Surveys; 2009; Yang, M., Falcone, A. E., Milan, L. M.
- Web based macroseismic survey: fast information exchange and elaboration of seismic intensity effects...; 2009; De Rubeis, V., Sbarra P., Sorrentino, D., Tosi, P.
- The representativeness of the LISS panel ; 2009; Knoef, M., de Vos, K.
- Sample factors that influence data quality; 2008; Gailey, R., Teal, D., Haechrel, E.
- An online panel as a platform for multi-disciplinary research; 2008; Scherpenzeel, A.
- Visual Design Effects on on Respondents Behaviour in Web-Surveys. A Design Experiment; 2008; Greinoecker, A.
- Effects of Privacy Assurances on the Online Measurement of Psychological Constructs; 2008; Witzki, A., Kramer, J.
- How Web 2.0 Technologies Can Become a Valuable Part of Online Research; 2008; Jaron, R.
- Respondent Authenticity - A biometrical approach to authenticate panelists; 2008; Wachter, B., Bender, C.
- Not Mixed-Mode but Switch-Mode; 2008; Höglinger, M., Abraham, M., Arpagaus, J.
- The Impact of Cognitive and Computer Skills on Data Quality in Computer Assisted Self Administered Questionnaires...; 2008; Brecko, B. N., Vehovar, V.
- Optimal Contact Strategy in a Mail-and-Web Mixed Mode Survey; 2008; Holmberg, A., Lorenc, B., Werner, P.
- 10 Years of Meinungsplatz.de: Success in the Collection of Data for Targeted Audiences, Such as the...; 2008; Weyergraf, O.
- Self-selection in Online Access Panels: No “Little Difference” in the Recruiting Process...; 2008; Wirth, T.
- Mobile Market Research; 2008; Maxl, E.
- Online vs. Offline in Mobile Surveys; 2008; Neubarth, W., Maier, U.
- Gender-of-Interviewer Effects in Video-Enhanced Web Surveys. Results from a Randomized Field-Experiment...; 2008; Fuchs, M.
- The Online Use of Randomized Response Measurements; 2008; Snijders, C., Weesie, J.